2025-04-01 21:59:42,882 [ 66926 ] INFO : ClickHouse root is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse (runner:53, check_args_and_update_paths) 2025-04-01 21:59:42,882 [ 66926 ] INFO : Cases dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:97, check_args_and_update_paths) 2025-04-01 21:59:42,882 [ 66926 ] INFO : utils dir is not set. Will use /home/ubuntu/_work/ClickHouse/ClickHouse/utils (runner:108, check_args_and_update_paths) 2025-04-01 21:59:42,882 [ 66926 ] INFO : base_configs_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/programs/server, binary: /home/ubuntu/_work/_temp/test/build/clickhouse, cases_dir: /home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration (runner:110, check_args_and_update_paths) clickhouse_integration_tests_volume Running pytest container as: 'docker run --rm --name clickhouse_integration_tests_bvh7ap --privileged --dns-search='.' --memory=30709035008 --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-odbc-bridge:/clickhouse-odbc-bridge --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-library-bridge:/clickhouse-library-bridge --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=6712d5cc610d -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=caad4729259e -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_accept_invalid_certificate/test.py::test_accept test_accept_invalid_certificate/test.py::test_connection_accept test_accept_invalid_certificate/test.py::test_default test_accept_invalid_certificate/test.py::test_strict_connection_reject test_accept_invalid_certificate/test.py::test_strict_reject test_accept_invalid_certificate/test.py::test_strict_reject_with_config test_always_fetch_merged/test.py::test_replica_always_download test_async_insert_memory/test.py::test_memory_usage test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field test_attach_partition_using_copy/test.py::test_all_replicated test_attach_partition_using_copy/test.py::test_both_mergetree test_attach_partition_using_copy/test.py::test_not_work_on_different_disk test_attach_partition_using_copy/test.py::test_only_destination_replicated test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication test_backup_restore/test.py::test_attach_partition test_backup_restore/test.py::test_replace_partition test_backup_restore/test.py::test_restore 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True]' test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes test_backup_restore_on_cluster/test.py::test_empty_replicated_table test_backup_restore_on_cluster/test.py::test_file_deduplication test_backup_restore_on_cluster/test.py::test_get_error_from_other_host test_backup_restore_on_cluster/test.py::test_keeper_value_max_size test_backup_restore_on_cluster/test.py::test_mutation test_backup_restore_on_cluster/test.py::test_projection test_backup_restore_on_cluster/test.py::test_replicated_database test_backup_restore_on_cluster/test.py::test_replicated_database_async test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database]' 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid]' test_backup_restore_on_cluster/test.py::test_replicated_table test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath test_backup_restore_on_cluster/test.py::test_required_privileges test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup test_backup_restore_on_cluster/test.py::test_system_functions test_backup_restore_on_cluster/test.py::test_system_users test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty test_backup_restore_on_cluster/test.py::test_tables_dependency test_cluster_all_replicas/test.py::test_cluster 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes]' test_cluster_all_replicas/test.py::test_global_in 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes]' test_compression_nested_columns/test.py::test_nested_compression_codec test_concurrent_queries_restriction_by_query_kind/test.py::test_insert test_concurrent_queries_restriction_by_query_kind/test.py::test_select test_config_decryption/test_wrong_settings.py::test_invalid_chars test_config_decryption/test_wrong_settings.py::test_no_encryption_key test_config_decryption/test_wrong_settings.py::test_subnodes test_config_decryption/test_wrong_settings.py::test_wrong_method test_config_xml_main/test.py::test_xml_main_conf test_config_yaml_main/test.py::test_yaml_main_conf test_create_query_constraints/test.py::test_create_query_const_constraints test_create_query_constraints/test.py::test_create_query_minmax_constraints test_custom_settings/test.py::test_custom_settings test_custom_settings/test.py::test_illformed_setting test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica test_ddl_worker_replicas/test.py::test_ddl_worker_replicas test_default_role/test.py::test_alter_user test_default_role/test.py::test_set_default_roles test_default_role/test.py::test_wrong_set_default_role 'test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed]' 'test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0]' 'test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1]' 'test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0]' 'test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1]' 'test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0]' 'test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1]' 'test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0]' 'test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1]' test_dictionaries_dependency/test.py::test_no_lazy_load 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed]' 'test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache]' 'test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache]' 'test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache]' test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation test_disk_access_storage/test.py::test_alter -vvv" altinityinfra/integration-tests-runner:cd6390247eca '. Start tests ============================= test session starts ============================== platform linux -- Python 3.10.12, pytest-7.4.4, pluggy-1.5.0 -- /usr/bin/python3 cachedir: .pytest_cache rootdir: /ClickHouse/tests/integration configfile: pytest.ini plugins: random-0.2, timeout-2.2.0, repeat-0.9.3, order-1.0.0, reportlog-0.4.0, xdist-3.5.0 timeout: 900.0s timeout method: signal timeout func_only: False created: 10/10 workers 10 workers [100 items] scheduling tests via LoadFileScheduling test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] test_default_role/test.py::test_alter_user test_accept_invalid_certificate/test.py::test_accept test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache] test_cluster_all_replicas/test.py::test_cluster test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0] test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed] test_attach_partition_using_copy/test.py::test_all_replicated test_backup_restore/test.py::test_attach_partition test_config_decryption/test_wrong_settings.py::test_invalid_chars [gw4] [ 1%] PASSED test_accept_invalid_certificate/test.py::test_accept test_accept_invalid_certificate/test.py::test_connection_accept [gw8] [ 2%] PASSED test_config_decryption/test_wrong_settings.py::test_invalid_chars test_config_decryption/test_wrong_settings.py::test_no_encryption_key [gw4] [ 3%] PASSED test_accept_invalid_certificate/test.py::test_connection_accept test_accept_invalid_certificate/test.py::test_default [gw4] [ 4%] PASSED test_accept_invalid_certificate/test.py::test_default test_accept_invalid_certificate/test.py::test_strict_connection_reject [gw4] [ 5%] PASSED test_accept_invalid_certificate/test.py::test_strict_connection_reject test_accept_invalid_certificate/test.py::test_strict_reject [gw4] [ 6%] PASSED test_accept_invalid_certificate/test.py::test_strict_reject test_accept_invalid_certificate/test.py::test_strict_reject_with_config [gw4] [ 7%] PASSED test_accept_invalid_certificate/test.py::test_strict_reject_with_config [gw1] [ 8%] PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed] test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed] [gw1] [ 9%] PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed] test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat] [gw1] [ 10%] PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat] test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed] [gw1] [ 11%] PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed] test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed] [gw9] [ 12%] PASSED test_default_role/test.py::test_alter_user test_default_role/test.py::test_set_default_roles [gw1] [ 13%] PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed] test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache] test_concurrent_queries_restriction_by_query_kind/test.py::test_insert [gw9] [ 14%] PASSED test_default_role/test.py::test_set_default_roles test_default_role/test.py::test_wrong_set_default_role [gw1] [ 15%] PASSED test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache] test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache] [gw7] [ 16%] PASSED test_backup_restore/test.py::test_attach_partition test_backup_restore/test.py::test_replace_partition [gw9] [ 17%] PASSED test_default_role/test.py::test_wrong_set_default_role [gw5] [ 18%] PASSED test_cluster_all_replicas/test.py::test_cluster test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] [gw0] [ 19%] PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] [gw1] [ 20%] PASSED test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache] test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache] test_create_query_constraints/test.py::test_create_query_const_constraints [gw8] [ 21%] PASSED test_config_decryption/test_wrong_settings.py::test_no_encryption_key test_config_decryption/test_wrong_settings.py::test_subnodes test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] [gw1] [ 22%] PASSED test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache] [gw5] [ 23%] PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] [gw7] [ 24%] PASSED test_backup_restore/test.py::test_replace_partition test_backup_restore/test.py::test_restore [gw0] [ 25%] PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache] test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] [gw5] [ 26%] PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] test_cluster_all_replicas/test.py::test_global_in [gw3] [ 27%] PASSED test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0] [gw7] [ 28%] PASSED test_backup_restore/test.py::test_restore [gw5] [ 29%] PASSED test_cluster_all_replicas/test.py::test_global_in test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] [gw0] [ 30%] PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] [gw8] [ 31%] PASSED test_config_decryption/test_wrong_settings.py::test_subnodes test_config_decryption/test_wrong_settings.py::test_wrong_method test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica [gw4] [ 32%] PASSED test_concurrent_queries_restriction_by_query_kind/test.py::test_insert test_concurrent_queries_restriction_by_query_kind/test.py::test_select test_always_fetch_merged/test.py::test_replica_always_download [gw9] [ 33%] PASSED test_create_query_constraints/test.py::test_create_query_const_constraints test_create_query_constraints/test.py::test_create_query_minmax_constraints test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1] [gw5] [ 34%] PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] [gw0] [ 35%] PASSED test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica [gw9] [ 36%] PASSED test_create_query_constraints/test.py::test_create_query_minmax_constraints test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes test_async_insert_memory/test.py::test_memory_usage [gw0] [ 37%] PASSED test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes test_backup_restore_on_cluster/test.py::test_empty_replicated_table [gw5] [ 38%] PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] [gw8] [ 39%] PASSED test_config_decryption/test_wrong_settings.py::test_wrong_method test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field [gw0] [ 40%] PASSED test_backup_restore_on_cluster/test.py::test_empty_replicated_table test_backup_restore_on_cluster/test.py::test_file_deduplication [gw2] [ 41%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache] test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct] [gw0] [ 42%] PASSED test_backup_restore_on_cluster/test.py::test_file_deduplication [gw4] [ 43%] PASSED test_concurrent_queries_restriction_by_query_kind/test.py::test_select test_backup_restore_on_cluster/test.py::test_get_error_from_other_host [gw3] [ 44%] PASSED test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1] test_config_xml_main/test.py::test_xml_main_conf [gw0] [ 45%] PASSED test_backup_restore_on_cluster/test.py::test_get_error_from_other_host test_backup_restore_on_cluster/test.py::test_keeper_value_max_size test_custom_settings/test.py::test_custom_settings test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0] [gw0] [ 46%] PASSED test_backup_restore_on_cluster/test.py::test_keeper_value_max_size [gw7] [ 47%] PASSED test_always_fetch_merged/test.py::test_replica_always_download test_backup_restore_on_cluster/test.py::test_mutation [gw8] [ 48%] PASSED test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field [gw4] [ 49%] PASSED test_custom_settings/test.py::test_custom_settings test_custom_settings/test.py::test_illformed_setting test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication [gw4] [ 50%] PASSED test_custom_settings/test.py::test_illformed_setting [gw0] [ 51%] PASSED test_backup_restore_on_cluster/test.py::test_mutation test_config_yaml_main/test.py::test_yaml_main_conf test_backup_restore_on_cluster/test.py::test_projection [gw3] [ 52%] PASSED test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0] [gw1] [ 53%] PASSED test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache] test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache] [gw5] [ 54%] PASSED test_config_xml_main/test.py::test_xml_main_conf test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation [gw0] [ 55%] PASSED test_backup_restore_on_cluster/test.py::test_projection test_backup_restore_on_cluster/test.py::test_replicated_database test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1] [gw0] [ 56%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database test_backup_restore_on_cluster/test.py::test_replicated_database_async [gw8] [ 57%] PASSED test_config_yaml_main/test.py::test_yaml_main_conf [gw3] [ 58%] PASSED test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1] [gw2] [ 59%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct] test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed] [gw5] [ 60%] PASSED test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation [gw0] [ 61%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_async test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0] [gw0] [ 62%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] [gw3] [ 63%] PASSED test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0] [gw0] [ 64%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1] [gw0] [ 65%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] [gw1] [ 66%] PASSED test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache] test_backup_restore_on_cluster/test.py::test_replicated_table test_compression_nested_columns/test.py::test_nested_compression_codec [gw0] [ 67%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table [gw3] [ 68%] PASSED test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1] test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster [gw0] [ 69%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0] test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster [gw2] [ 70%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed] test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed] [gw0] [ 71%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def [gw0] [ 72%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert [gw1] [ 73%] PASSED test_compression_nested_columns/test.py::test_nested_compression_codec [gw3] [ 74%] PASSED test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0] [gw0] [ 75%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge [gw2] [ 76%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed] test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache] test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1] [gw0] [ 77%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath [gw0] [ 78%] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath test_backup_restore_on_cluster/test.py::test_required_privileges [gw0] [ 79%] PASSED test_backup_restore_on_cluster/test.py::test_required_privileges test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup [gw3] [ 80%] PASSED test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1] test_dictionaries_dependency/test.py::test_no_lazy_load [gw0] [ 81%] PASSED test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup test_backup_restore_on_cluster/test.py::test_system_functions [gw2] [ 82%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache] test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct] [gw0] [ 83%] PASSED test_backup_restore_on_cluster/test.py::test_system_functions test_backup_restore_on_cluster/test.py::test_system_users [gw0] [ 84%] PASSED test_backup_restore_on_cluster/test.py::test_system_users test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def [gw0] [ 85%] PASSED test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty [gw0] [ 86%] PASSED test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty test_backup_restore_on_cluster/test.py::test_tables_dependency [gw0] [ 87%] PASSED test_backup_restore_on_cluster/test.py::test_tables_dependency [gw3] [ 88%] PASSED test_dictionaries_dependency/test.py::test_no_lazy_load [gw2] [ 89%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct] test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat] [gw7] [ 90%] PASSED test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication [gw9] [ 91%] PASSED test_async_insert_memory/test.py::test_memory_usage test_ddl_worker_replicas/test.py::test_ddl_worker_replicas test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica [gw7] [ 92%] PASSED test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica [gw2] [ 93%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat] test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed] [gw2] [ 94%] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed] [gw9] [ 95%] PASSED test_ddl_worker_replicas/test.py::test_ddl_worker_replicas test_disk_access_storage/test.py::test_alter [gw9] [ 96%] PASSED test_disk_access_storage/test.py::test_alter [gw6] [ 97%] FAILED test_attach_partition_using_copy/test.py::test_all_replicated test_attach_partition_using_copy/test.py::test_both_mergetree [gw6] [ 98%] FAILED test_attach_partition_using_copy/test.py::test_both_mergetree test_attach_partition_using_copy/test.py::test_not_work_on_different_disk [gw6] [ 99%] FAILED test_attach_partition_using_copy/test.py::test_not_work_on_different_disk test_attach_partition_using_copy/test.py::test_only_destination_replicated [gw6] [100%] FAILED test_attach_partition_using_copy/test.py::test_only_destination_replicated =================================== FAILURES =================================== _____________________________ test_all_replicated ______________________________ [gw6] linux -- Python 3.10.12 /usr/bin/python3 start_cluster = def test_all_replicated(start_cluster): cleanup([replica1, replica2]) > create_source_table(replica1, "source", True) test_attach_partition_using_copy/test.py:126: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( helpers/cluster.py:3713: in query_with_retry result = self.query( helpers/cluster.py:3678: in query return self.client.query( helpers/client.py:39: in wrap return func(self, *args, **kwargs) helpers/client.py:79: in query ).get_answer() helpers/client.py:230: in get_answer self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) /usr/lib/python3.10/subprocess.py:1209: in wait return self._wait(timeout=timeout) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = timeout = 600 def _wait(self, timeout): """Internal implementation of wait() on POSIX.""" if self.returncode is not None: return self.returncode if timeout is not None: endtime = _time() + timeout # Enter a busy loop if we have a timeout. This busy loop was # cribbed from Lib/threading.py in Thread.wait() at r71065. delay = 0.0005 # 500 us -> initial delay of 1 ms while True: if self._waitpid_lock.acquire(False): try: if self.returncode is not None: break # Another thread waited. (pid, sts) = self._try_wait(os.WNOHANG) assert pid == self.pid or pid == 0 if pid == self.pid: self._handle_exitstatus(sts) break finally: self._waitpid_lock.release() remaining = self._remaining_time(endtime) if remaining <= 0: raise TimeoutExpired(self.args, timeout) delay = min(delay * 2, remaining, .05) > time.sleep(delay) E Failed: Timeout >900.0s /usr/lib/python3.10/subprocess.py:1953: Failed ---------------------------- Captured stdout setup ----------------------------- Copy common default production configuration from /clickhouse-config. Files: config.xml, users.xml Copy common default production configuration from /clickhouse-config. Files: config.xml, users.xml ---------------------------- Captured stderr setup ----------------------------- Command:[docker ps | wc -l] Stdout:1 No running containers Pruning Docker networks Command:[docker network prune --force] Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] Stdout:net.ipv4.ip_local_port_range = 55000 65535 Running tests in /ClickHouse/tests/integration/test_attach_partition_using_copy/test.py Cluster start called. is_up=False Docker networks for project roottestattachpartitionusingcopy-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestattachpartitionusingcopy-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestattachpartitionusingcopy-gw6 are DRIVER VOLUME NAME Cleanup called Docker networks for project roottestattachpartitionusingcopy-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestattachpartitionusingcopy-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestattachpartitionusingcopy-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestattachpartitionusingcopy-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestattachpartitionusingcopy-gw6 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 Setup directory for instance: replica1 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_attach_partition_using_copy/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/configs/config.d Setup database dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/database Setup logs dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Setup directory for instance: replica2 Create directory for configuration generated in this helper Create directory for common tests configuration Copy common configuration from helpers Generate and write macros file Copy custom test config files ['/ClickHouse/tests/integration/test_attach_partition_using_copy/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/configs/config.d Setup database dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/database Setup logs dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/logs Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:6712d5cc610d', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] No config file found http://localhost:None "GET /version HTTP/1.1" 200 826 Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml pull] Stderr: zoo1 Skipped - Image is already being pulled by zoo3 Stderr: replica2 Skipped - Image is already being pulled by zoo3 Stderr: replica1 Skipped - Image is already being pulled by zoo3 Stderr: zoo2 Skipped - Image is already being pulled by zoo3 Stderr: zoo3 Pulling Stderr: zoo3 Pulled Setup ZooKeeper Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/coordination', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/coordination', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/coordination'] Command:[docker compose --project-name roottestattachpartitionusingcopy-gw6 --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] Stderr:time="2025-04-01T22:00:00Z" level=trace msg="Docker Desktop integration not enabled" Stderr: Network roottestattachpartitionusingcopy-gw6_default Creating Stderr: Network roottestattachpartitionusingcopy-gw6_default Created Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Creating Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Creating Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Creating Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Created Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Created Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Created Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Starting Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Starting Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Starting Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Started Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Started Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Started Stderr:time="2025-04-01T22:00:02Z" level=debug msg="otel error" error="" Stderr:time="2025-04-01T22:00:02Z" level=debug msg="otel error" error="" Wait ZooKeeper to start get_instance_ip instance_name=zoo1 http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-zoo1-1/json HTTP/1.1" 200 None get_kazoo_client: zoo1, ip:172.16.6.4, port:2181, use_ssl:False Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Connection dropped: socket connection error: Connection refused Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=10000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo2 http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-zoo2-1/json HTTP/1.1" 200 None get_kazoo_client: zoo2, ip:172.16.6.2, port:2181, use_ssl:False Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=10000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED get_instance_ip instance_name=zoo3 http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-zoo3-1/json HTTP/1.1" 200 None get_kazoo_client: zoo3, ip:172.16.6.3, port:2181, use_ssl:False Connecting to 172.16.6.3(172.16.6.3):2181, use_ssl: False Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=10000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) Zookeeper connection established, state: CONNECTED Sending request(xid=1): GetChildren(path='/', watcher=None) Received response(xid=1): ['keeper'] Sending request(xid=2): Close() Connection dropped: socket connection broken Transition to CONNECTING Zookeeper connection lost Failed connecting to Zookeeper within the connection retry policy. Zookeeper session closed, state: CLOSED All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml up -d --no-recreate') Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml up -d --no-recreate] Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Running Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Running Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Running Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Creating Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Creating Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Created Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Created Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Starting Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Starting Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Started Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Started ClickHouse instance created get_instance_ip instance_name=replica1 http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica1-1/json HTTP/1.1" 200 None get_instance_ip instance_name=replica1 http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica1-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in replica1, ip: 172.16.6.6... http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica1-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None ClickHouse replica1 started get_instance_ip instance_name=replica2 http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica2-1/json HTTP/1.1" 200 None get_instance_ip instance_name=replica2 http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica2-1/json HTTP/1.1" 200 None Waiting for ClickHouse start in replica2, ip: 172.16.6.5... http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica2-1/json HTTP/1.1" 200 None http://localhost:None "GET /v1.46/containers/e10d2fe27b53a2090068e8e210aaa647383db37aac0017de405382d17797cd70/json HTTP/1.1" 200 None ClickHouse replica2 started ------------------------------ Captured log setup ------------------------------ 2025-04-01 21:59:49 [ 694 ] DEBUG : Command:[docker ps | wc -l] (cluster.py:122, run_and_check) 2025-04-01 21:59:49 [ 694 ] DEBUG : Stdout:1 (cluster.py:146, run_and_check) 2025-04-01 21:59:49 [ 694 ] DEBUG : No running containers (conftest.py:96, cleanup_environment) 2025-04-01 21:59:49 [ 694 ] DEBUG : Pruning Docker networks (conftest.py:98, cleanup_environment) 2025-04-01 21:59:49 [ 694 ] DEBUG : Command:[docker network prune --force] (cluster.py:122, run_and_check) 2025-04-01 21:59:49 [ 694 ] DEBUG : Command:[sysctl net.ipv4.ip_local_port_range='55000 65535'] (cluster.py:122, run_and_check) 2025-04-01 21:59:49 [ 694 ] DEBUG : Stdout:net.ipv4.ip_local_port_range = 55000 65535 (cluster.py:146, run_and_check) 2025-04-01 21:59:49 [ 694 ] INFO : Running tests in /ClickHouse/tests/integration/test_attach_partition_using_copy/test.py (cluster.py:2793, start) 2025-04-01 21:59:49 [ 694 ] DEBUG : Cluster start called. is_up=False (cluster.py:2800, start) 2025-04-01 21:59:49 [ 694 ] DEBUG : Docker networks for project roottestattachpartitionusingcopy-gw6 are NETWORK ID NAME DRIVER SCOPE (cluster.py:873, print_all_docker_pieces) 2025-04-01 21:59:49 [ 694 ] DEBUG : Docker containers for project roottestattachpartitionusingcopy-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:881, print_all_docker_pieces) 2025-04-01 21:59:49 [ 694 ] DEBUG : Docker volumes for project roottestattachpartitionusingcopy-gw6 are DRIVER VOLUME NAME (cluster.py:889, print_all_docker_pieces) 2025-04-01 21:59:49 [ 694 ] DEBUG : Cleanup called (cluster.py:894, cleanup) 2025-04-01 21:59:49 [ 694 ] DEBUG : Docker networks for project roottestattachpartitionusingcopy-gw6 are NETWORK ID NAME DRIVER SCOPE (cluster.py:873, print_all_docker_pieces) 2025-04-01 21:59:49 [ 694 ] DEBUG : Docker containers for project roottestattachpartitionusingcopy-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:881, print_all_docker_pieces) 2025-04-01 21:59:49 [ 694 ] DEBUG : Docker volumes for project roottestattachpartitionusingcopy-gw6 are DRIVER VOLUME NAME (cluster.py:889, print_all_docker_pieces) 2025-04-01 21:59:49 [ 694 ] DEBUG : Command:[docker container list --all --filter name='^/roottestattachpartitionusingcopy-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:122, run_and_check) 2025-04-01 21:59:49 [ 694 ] DEBUG : Unstopped containers: {} (cluster.py:908, cleanup) 2025-04-01 21:59:49 [ 694 ] DEBUG : No running containers for project: roottestattachpartitionusingcopy-gw6 (cluster.py:922, cleanup) 2025-04-01 21:59:49 [ 694 ] DEBUG : Trying to prune unused networks... (cluster.py:928, cleanup) 2025-04-01 21:59:49 [ 694 ] DEBUG : Trying to prune unused images... (cluster.py:944, cleanup) 2025-04-01 21:59:49 [ 694 ] DEBUG : Command:[docker image prune -f] (cluster.py:122, run_and_check) 2025-04-01 21:59:49 [ 694 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:146, run_and_check) 2025-04-01 21:59:49 [ 694 ] DEBUG : Images pruned (cluster.py:947, cleanup) 2025-04-01 21:59:49 [ 694 ] DEBUG : Trying to prune unused volumes... (cluster.py:953, cleanup) 2025-04-01 21:59:49 [ 694 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:122, run_and_check) 2025-04-01 21:59:49 [ 694 ] DEBUG : Stdout:1 (cluster.py:146, run_and_check) 2025-04-01 21:59:49 [ 694 ] DEBUG : Volumes pruned: 1 (cluster.py:958, cleanup) 2025-04-01 21:59:49 [ 694 ] DEBUG : Setup directory for instance: replica1 (cluster.py:2813, start) 2025-04-01 21:59:49 [ 694 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4639, create_dir) 2025-04-01 21:59:49 [ 694 ] DEBUG : Create directory for common tests configuration (cluster.py:4644, create_dir) 2025-04-01 21:59:49 [ 694 ] DEBUG : Copy common configuration from helpers (cluster.py:4664, create_dir) 2025-04-01 21:59:49 [ 694 ] DEBUG : Generate and write macros file (cluster.py:4716, create_dir) 2025-04-01 21:59:49 [ 694 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_attach_partition_using_copy/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/configs/config.d (cluster.py:4752, create_dir) 2025-04-01 21:59:49 [ 694 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/database (cluster.py:4769, create_dir) 2025-04-01 21:59:49 [ 694 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/logs (cluster.py:4780, create_dir) 2025-04-01 21:59:49 [ 694 ] DEBUG : Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] (cluster.py:4864, create_dir) 2025-04-01 21:59:49 [ 694 ] DEBUG : Setup directory for instance: replica2 (cluster.py:2813, start) 2025-04-01 21:59:49 [ 694 ] DEBUG : Create directory for configuration generated in this helper (cluster.py:4639, create_dir) 2025-04-01 21:59:49 [ 694 ] DEBUG : Create directory for common tests configuration (cluster.py:4644, create_dir) 2025-04-01 21:59:49 [ 694 ] DEBUG : Copy common configuration from helpers (cluster.py:4664, create_dir) 2025-04-01 21:59:49 [ 694 ] DEBUG : Generate and write macros file (cluster.py:4716, create_dir) 2025-04-01 21:59:49 [ 694 ] DEBUG : Copy custom test config files ['/ClickHouse/tests/integration/test_attach_partition_using_copy/configs/remote_servers.xml'] to /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/configs/config.d (cluster.py:4752, create_dir) 2025-04-01 21:59:49 [ 694 ] DEBUG : Setup database dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/database (cluster.py:4769, create_dir) 2025-04-01 21:59:49 [ 694 ] DEBUG : Setup logs dir /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/logs (cluster.py:4780, create_dir) 2025-04-01 21:59:49 [ 694 ] DEBUG : Entrypoint cmd: ["clickhouse", "server", "--config-file=/etc/clickhouse-server/config.xml", "--log-file=/var/log/clickhouse-server/clickhouse-server.log", "--errorlog-file=/var/log/clickhouse-server/clickhouse-server.err.log", "--"] (cluster.py:4864, create_dir) 2025-04-01 21:59:49 [ 694 ] DEBUG : Env {'ASAN_OPTIONS': 'use_sigaltstack=0', 'TSAN_OPTIONS': 'use_sigaltstack=0', 'CLICKHOUSE_WATCHDOG_ENABLE': '0', 'CLICKHOUSE_NATS_TLS_SECURE': '0', 'LLVM_PROFILE_FILE': '/var/lib/clickhouse/server_%h_%p_%m.profraw', 'keeper_binary': '/clickhouse', 'keeper_cmd_prefix': 'clickhouse keeper', 'image': 'altinityinfra/integration-test:6712d5cc610d', 'user': '0', 'keeper_fs': 'bind', 'keeper_logs_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/log', 'keeper_config_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/config', 'keeper_db_dir1': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/coordination', 'keeper_logs_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/log', 'keeper_config_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/config', 'keeper_db_dir2': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/coordination', 'keeper_logs_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/log', 'keeper_config_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/config', 'keeper_db_dir3': '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/coordination'} stored in /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env (cluster.py:97, _create_env_file) 2025-04-01 21:59:49 [ 694 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2025-04-01 21:59:49 [ 694 ] DEBUG : No config file found (config.py:28, find_config_file) 2025-04-01 21:59:49 [ 694 ] DEBUG : Trying paths: ['/root/.docker/config.json', '/root/.dockercfg'] (config.py:21, find_config_file) 2025-04-01 21:59:49 [ 694 ] DEBUG : No config file found (config.py:28, find_config_file) 2025-04-01 21:59:49 [ 694 ] DEBUG : http://localhost:None "GET /version HTTP/1.1" 200 826 (connectionpool.py:547, _make_request) 2025-04-01 21:59:49 [ 694 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml pull] (cluster.py:122, run_and_check) 2025-04-01 22:00:00 [ 694 ] DEBUG : Stderr: zoo1 Skipped - Image is already being pulled by zoo3 (cluster.py:148, run_and_check) 2025-04-01 22:00:00 [ 694 ] DEBUG : Stderr: replica2 Skipped - Image is already being pulled by zoo3 (cluster.py:148, run_and_check) 2025-04-01 22:00:00 [ 694 ] DEBUG : Stderr: replica1 Skipped - Image is already being pulled by zoo3 (cluster.py:148, run_and_check) 2025-04-01 22:00:00 [ 694 ] DEBUG : Stderr: zoo2 Skipped - Image is already being pulled by zoo3 (cluster.py:148, run_and_check) 2025-04-01 22:00:00 [ 694 ] DEBUG : Stderr: zoo3 Pulling (cluster.py:148, run_and_check) 2025-04-01 22:00:00 [ 694 ] DEBUG : Stderr: zoo3 Pulled (cluster.py:148, run_and_check) 2025-04-01 22:00:00 [ 694 ] DEBUG : Setup ZooKeeper (cluster.py:2854, start) 2025-04-01 22:00:00 [ 694 ] DEBUG : Creating internal ZooKeeper dirs: ['/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper1/coordination', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper2/coordination', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/log', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/config', '/ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/keeper3/coordination'] (cluster.py:2855, start) 2025-04-01 22:00:00 [ 694 ] DEBUG : Command:[docker compose --project-name roottestattachpartitionusingcopy-gw6 --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --verbose up -d] (cluster.py:122, run_and_check) 2025-04-01 22:00:02 [ 694 ] DEBUG : Stderr:time="2025-04-01T22:00:00Z" level=trace msg="Docker Desktop integration not enabled" (cluster.py:148, run_and_check) 2025-04-01 22:00:02 [ 694 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy-gw6_default Creating (cluster.py:148, run_and_check) 2025-04-01 22:00:02 [ 694 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy-gw6_default Created (cluster.py:148, run_and_check) 2025-04-01 22:00:02 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Creating (cluster.py:148, run_and_check) 2025-04-01 22:00:02 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Creating (cluster.py:148, run_and_check) 2025-04-01 22:00:02 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Creating (cluster.py:148, run_and_check) 2025-04-01 22:00:02 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Created (cluster.py:148, run_and_check) 2025-04-01 22:00:02 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Created (cluster.py:148, run_and_check) 2025-04-01 22:00:02 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Created (cluster.py:148, run_and_check) 2025-04-01 22:00:02 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Starting (cluster.py:148, run_and_check) 2025-04-01 22:00:02 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Starting (cluster.py:148, run_and_check) 2025-04-01 22:00:02 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Starting (cluster.py:148, run_and_check) 2025-04-01 22:00:02 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Started (cluster.py:148, run_and_check) 2025-04-01 22:00:02 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Started (cluster.py:148, run_and_check) 2025-04-01 22:00:02 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Started (cluster.py:148, run_and_check) 2025-04-01 22:00:02 [ 694 ] DEBUG : Stderr:time="2025-04-01T22:00:02Z" level=debug msg="otel error" error="" (cluster.py:148, run_and_check) 2025-04-01 22:00:02 [ 694 ] DEBUG : Stderr:time="2025-04-01T22:00:02Z" level=debug msg="otel error" error="" (cluster.py:148, run_and_check) 2025-04-01 22:00:02 [ 694 ] DEBUG : Wait ZooKeeper to start (cluster.py:2466, wait_zookeeper_to_start) 2025-04-01 22:00:02 [ 694 ] DEBUG : get_instance_ip instance_name=zoo1 (cluster.py:2082, get_instance_ip) 2025-04-01 22:00:02 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-zoo1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:02 [ 694 ] DEBUG : get_kazoo_client: zoo1, ip:172.16.6.4, port:2181, use_ssl:False (cluster.py:3341, get_kazoo_client) 2025-04-01 22:00:02 [ 694 ] INFO : Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False (connection.py:650, _connect) 2025-04-01 22:00:02 [ 694 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-01 22:00:02 [ 694 ] INFO : Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False (connection.py:650, _connect) 2025-04-01 22:00:02 [ 694 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-01 22:00:02 [ 694 ] INFO : Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False (connection.py:650, _connect) 2025-04-01 22:00:02 [ 694 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-01 22:00:02 [ 694 ] INFO : Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False (connection.py:650, _connect) 2025-04-01 22:00:02 [ 694 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-01 22:00:03 [ 694 ] INFO : Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False (connection.py:650, _connect) 2025-04-01 22:00:03 [ 694 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-01 22:00:04 [ 694 ] INFO : Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False (connection.py:650, _connect) 2025-04-01 22:00:04 [ 694 ] WARNING : Connection dropped: socket connection error: Connection refused (connection.py:622, _connect_attempt) 2025-04-01 22:00:07 [ 694 ] INFO : Connecting to 172.16.6.4(172.16.6.4):2181, use_ssl: False (connection.py:650, _connect) 2025-04-01 22:00:07 [ 694 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=10000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2025-04-01 22:00:07 [ 694 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2025-04-01 22:00:07 [ 694 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2025-04-01 22:00:07 [ 694 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2025-04-01 22:00:07 [ 694 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2025-04-01 22:00:07 [ 694 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2025-04-01 22:00:07 [ 694 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2025-04-01 22:00:07 [ 694 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2025-04-01 22:00:07 [ 694 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2025-04-01 22:00:07 [ 694 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2025-04-01 22:00:07 [ 694 ] DEBUG : get_instance_ip instance_name=zoo2 (cluster.py:2082, get_instance_ip) 2025-04-01 22:00:07 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-zoo2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:07 [ 694 ] DEBUG : get_kazoo_client: zoo2, ip:172.16.6.2, port:2181, use_ssl:False (cluster.py:3341, get_kazoo_client) 2025-04-01 22:00:07 [ 694 ] INFO : Connecting to 172.16.6.2(172.16.6.2):2181, use_ssl: False (connection.py:650, _connect) 2025-04-01 22:00:07 [ 694 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=10000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2025-04-01 22:00:07 [ 694 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2025-04-01 22:00:07 [ 694 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2025-04-01 22:00:07 [ 694 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2025-04-01 22:00:07 [ 694 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2025-04-01 22:00:07 [ 694 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2025-04-01 22:00:07 [ 694 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2025-04-01 22:00:07 [ 694 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2025-04-01 22:00:07 [ 694 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2025-04-01 22:00:07 [ 694 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2025-04-01 22:00:07 [ 694 ] DEBUG : get_instance_ip instance_name=zoo3 (cluster.py:2082, get_instance_ip) 2025-04-01 22:00:07 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-zoo3-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:07 [ 694 ] DEBUG : get_kazoo_client: zoo3, ip:172.16.6.3, port:2181, use_ssl:False (cluster.py:3341, get_kazoo_client) 2025-04-01 22:00:07 [ 694 ] INFO : Connecting to 172.16.6.3(172.16.6.3):2181, use_ssl: False (connection.py:650, _connect) 2025-04-01 22:00:07 [ 694 ] DEBUG : Sending request(xid=None): Connect(protocol_version=0, last_zxid_seen=0, time_out=10000, session_id=0, passwd=b'\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00\x00', read_only=None) (connection.py:312, _submit) 2025-04-01 22:00:07 [ 694 ] INFO : Zookeeper connection established, state: CONNECTED (client.py:532, _session_callback) 2025-04-01 22:00:07 [ 694 ] DEBUG : Sending request(xid=1): GetChildren(path='/', watcher=None) (connection.py:312, _submit) 2025-04-01 22:00:07 [ 694 ] DEBUG : Received response(xid=1): ['keeper'] (connection.py:410, _read_response) 2025-04-01 22:00:07 [ 694 ] DEBUG : Sending request(xid=2): Close() (connection.py:312, _submit) 2025-04-01 22:00:07 [ 694 ] WARNING : Connection dropped: socket connection broken (connection.py:622, _connect_attempt) 2025-04-01 22:00:07 [ 694 ] WARNING : Transition to CONNECTING (connection.py:626, _connect_attempt) 2025-04-01 22:00:07 [ 694 ] INFO : Zookeeper connection lost (client.py:543, _session_callback) 2025-04-01 22:00:07 [ 694 ] WARNING : Failed connecting to Zookeeper within the connection retry policy. (connection.py:515, zk_loop) 2025-04-01 22:00:07 [ 694 ] INFO : Zookeeper session closed, state: CLOSED (client.py:537, _session_callback) 2025-04-01 22:00:07 [ 694 ] DEBUG : All instances of ZooKeeper started: ('zoo1', 'zoo2', 'zoo3') (cluster.py:2482, wait_zookeeper_nodes_to_start) 2025-04-01 22:00:07 [ 694 ] DEBUG : ('Trying to create ClickHouse instance by command %s', 'docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml up -d --no-recreate') (cluster.py:3200, start) 2025-04-01 22:00:07 [ 694 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml up -d --no-recreate] (cluster.py:122, run_and_check) 2025-04-01 22:00:08 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Running (cluster.py:148, run_and_check) 2025-04-01 22:00:08 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Running (cluster.py:148, run_and_check) 2025-04-01 22:00:08 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Running (cluster.py:148, run_and_check) 2025-04-01 22:00:08 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Creating (cluster.py:148, run_and_check) 2025-04-01 22:00:08 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Creating (cluster.py:148, run_and_check) 2025-04-01 22:00:08 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Created (cluster.py:148, run_and_check) 2025-04-01 22:00:08 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Created (cluster.py:148, run_and_check) 2025-04-01 22:00:08 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Starting (cluster.py:148, run_and_check) 2025-04-01 22:00:08 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Starting (cluster.py:148, run_and_check) 2025-04-01 22:00:08 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Started (cluster.py:148, run_and_check) 2025-04-01 22:00:08 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Started (cluster.py:148, run_and_check) 2025-04-01 22:00:08 [ 694 ] DEBUG : ClickHouse instance created (cluster.py:3208, start) 2025-04-01 22:00:08 [ 694 ] DEBUG : get_instance_ip instance_name=replica1 (cluster.py:2082, get_instance_ip) 2025-04-01 22:00:08 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:08 [ 694 ] DEBUG : get_instance_ip instance_name=replica1 (cluster.py:2092, get_instance_global_ipv6) 2025-04-01 22:00:08 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:08 [ 694 ] DEBUG : Waiting for ClickHouse start in replica1, ip: 172.16.6.6... (cluster.py:3216, start) 2025-04-01 22:00:08 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica1-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:08 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:08 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:08 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:08 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:08 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:09 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:09 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:09 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:09 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:09 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:09 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:09 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:09 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:09 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:10 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:10 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:10 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:10 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:10 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:10 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:10 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:10 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:10 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:10 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/6f0f0d65ad1eb3d2fbf6cdb8a6e8db1b67d6e30903eafe842cb841c2c415933d/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:10 [ 694 ] DEBUG : ClickHouse replica1 started (cluster.py:3220, start) 2025-04-01 22:00:10 [ 694 ] DEBUG : get_instance_ip instance_name=replica2 (cluster.py:2082, get_instance_ip) 2025-04-01 22:00:10 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:10 [ 694 ] DEBUG : get_instance_ip instance_name=replica2 (cluster.py:2092, get_instance_global_ipv6) 2025-04-01 22:00:10 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:10 [ 694 ] DEBUG : Waiting for ClickHouse start in replica2, ip: 172.16.6.5... (cluster.py:3216, start) 2025-04-01 22:00:10 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/roottestattachpartitionusingcopy-gw6-replica2-1/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:10 [ 694 ] DEBUG : http://localhost:None "GET /v1.46/containers/e10d2fe27b53a2090068e8e210aaa647383db37aac0017de405382d17797cd70/json HTTP/1.1" 200 None (connectionpool.py:547, _make_request) 2025-04-01 22:00:10 [ 694 ] DEBUG : ClickHouse replica2 started (cluster.py:3220, start) ----------------------------- Captured stderr call ----------------------------- Executing query DROP TABLE IF EXISTS source SYNC on replica1 Executing query DROP TABLE IF EXISTS destination SYNC on replica1 Executing query DROP TABLE IF EXISTS source SYNC on replica2 Executing query DROP TABLE IF EXISTS destination SYNC on replica2 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 ~~~~~~~~~~~~~~~~~~~~~ Stack of (139747784914496) ~~~~~~~~~~~~~~~~~~~~~ File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 411, in _perform_spawn reply.run() File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 341, in run self._result = func(*args, **kwargs) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 1160, in _thread_receiver msg = Message.from_io(io) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 567, in from_io header = io.read(9) # type 1, channel 4, payload 4 File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 534, in read data = self._read(numbytes - len(buf)) ------------------------------ Captured log call ------------------------------- 2025-04-01 22:00:10 [ 694 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3677, query) 2025-04-01 22:00:11 [ 694 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3677, query) 2025-04-01 22:00:11 [ 694 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3677, query) 2025-04-01 22:00:11 [ 694 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3677, query) 2025-04-01 22:00:12 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:01:05 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:02:00 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:02:55 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:03:52 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:04:50 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:05:45 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:06:40 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:07:37 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:08:35 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:09:29 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:10:25 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:11:22 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:12:20 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:13:14 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:14:09 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = ReplicatedMergeTree('/clickhouse/tables/1/source', 'replica1') ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) _____________________________ test_both_mergetree ______________________________ [gw6] linux -- Python 3.10.12 /usr/bin/python3 start_cluster = def test_both_mergetree(start_cluster): cleanup([replica1, replica2]) > create_source_table(replica1, "source", False) test_attach_partition_using_copy/test.py:104: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( helpers/cluster.py:3713: in query_with_retry result = self.query( helpers/cluster.py:3678: in query return self.client.query( helpers/client.py:39: in wrap return func(self, *args, **kwargs) helpers/client.py:79: in query ).get_answer() helpers/client.py:230: in get_answer self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) /usr/lib/python3.10/subprocess.py:1209: in wait return self._wait(timeout=timeout) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = timeout = 600 def _wait(self, timeout): """Internal implementation of wait() on POSIX.""" if self.returncode is not None: return self.returncode if timeout is not None: endtime = _time() + timeout # Enter a busy loop if we have a timeout. This busy loop was # cribbed from Lib/threading.py in Thread.wait() at r71065. delay = 0.0005 # 500 us -> initial delay of 1 ms while True: if self._waitpid_lock.acquire(False): try: if self.returncode is not None: break # Another thread waited. (pid, sts) = self._try_wait(os.WNOHANG) assert pid == self.pid or pid == 0 if pid == self.pid: self._handle_exitstatus(sts) break finally: self._waitpid_lock.release() remaining = self._remaining_time(endtime) if remaining <= 0: raise TimeoutExpired(self.args, timeout) delay = min(delay * 2, remaining, .05) > time.sleep(delay) E Failed: Timeout >900.0s /usr/lib/python3.10/subprocess.py:1953: Failed ----------------------------- Captured stderr call ----------------------------- Executing query DROP TABLE IF EXISTS source SYNC on replica1 Executing query DROP TABLE IF EXISTS destination SYNC on replica1 Executing query DROP TABLE IF EXISTS source SYNC on replica2 Executing query DROP TABLE IF EXISTS destination SYNC on replica2 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 ~~~~~~~~~~~~~~~~~~~~~ Stack of (139747784914496) ~~~~~~~~~~~~~~~~~~~~~ File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 411, in _perform_spawn reply.run() File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 341, in run self._result = func(*args, **kwargs) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 1160, in _thread_receiver msg = Message.from_io(io) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 567, in from_io header = io.read(9) # type 1, channel 4, payload 4 File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 534, in read data = self._read(numbytes - len(buf)) ------------------------------ Captured log call ------------------------------- 2025-04-01 22:14:49 [ 694 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3677, query) 2025-04-01 22:15:06 [ 694 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3677, query) 2025-04-01 22:15:06 [ 694 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3677, query) 2025-04-01 22:15:06 [ 694 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3677, query) 2025-04-01 22:15:07 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:16:04 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:16:59 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:17:54 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:18:51 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:19:49 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:20:44 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:21:39 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:22:36 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:23:34 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:24:28 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:25:24 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:26:21 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:27:18 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:28:13 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:29:08 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) _______________________ test_not_work_on_different_disk ________________________ [gw6] linux -- Python 3.10.12 /usr/bin/python3 start_cluster = def test_not_work_on_different_disk(start_cluster): cleanup([replica1, replica2]) # Replace and move should not work on replace > create_source_table(replica1, "source", False) test_attach_partition_using_copy/test.py:197: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( helpers/cluster.py:3713: in query_with_retry result = self.query( helpers/cluster.py:3678: in query return self.client.query( helpers/client.py:39: in wrap return func(self, *args, **kwargs) helpers/client.py:79: in query ).get_answer() helpers/client.py:230: in get_answer self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) /usr/lib/python3.10/subprocess.py:1209: in wait return self._wait(timeout=timeout) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = timeout = 600 def _wait(self, timeout): """Internal implementation of wait() on POSIX.""" if self.returncode is not None: return self.returncode if timeout is not None: endtime = _time() + timeout # Enter a busy loop if we have a timeout. This busy loop was # cribbed from Lib/threading.py in Thread.wait() at r71065. delay = 0.0005 # 500 us -> initial delay of 1 ms while True: if self._waitpid_lock.acquire(False): try: if self.returncode is not None: break # Another thread waited. (pid, sts) = self._try_wait(os.WNOHANG) assert pid == self.pid or pid == 0 if pid == self.pid: self._handle_exitstatus(sts) break finally: self._waitpid_lock.release() remaining = self._remaining_time(endtime) if remaining <= 0: raise TimeoutExpired(self.args, timeout) delay = min(delay * 2, remaining, .05) > time.sleep(delay) E Failed: Timeout >900.0s /usr/lib/python3.10/subprocess.py:1953: Failed ----------------------------- Captured stderr call ----------------------------- Executing query DROP TABLE IF EXISTS source SYNC on replica1 Executing query DROP TABLE IF EXISTS destination SYNC on replica1 Executing query DROP TABLE IF EXISTS source SYNC on replica2 Executing query DROP TABLE IF EXISTS destination SYNC on replica2 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 ~~~~~~~~~~~~~~~~~~~~~ Stack of (139747784914496) ~~~~~~~~~~~~~~~~~~~~~ File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 411, in _perform_spawn reply.run() File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 341, in run self._result = func(*args, **kwargs) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 1160, in _thread_receiver msg = Message.from_io(io) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 567, in from_io header = io.read(9) # type 1, channel 4, payload 4 File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 534, in read data = self._read(numbytes - len(buf)) ------------------------------ Captured log call ------------------------------- 2025-04-01 22:29:49 [ 694 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3677, query) 2025-04-01 22:30:05 [ 694 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3677, query) 2025-04-01 22:30:05 [ 694 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3677, query) 2025-04-01 22:30:05 [ 694 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3677, query) 2025-04-01 22:30:05 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:31:03 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:31:58 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:32:53 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:33:50 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:34:48 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:35:43 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:36:38 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:37:35 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:38:33 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:39:27 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:40:22 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:41:19 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:42:17 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:43:12 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:44:07 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) _______________________ test_only_destination_replicated _______________________ [gw6] linux -- Python 3.10.12 /usr/bin/python3 start_cluster = def test_only_destination_replicated(start_cluster): cleanup([replica1, replica2]) > create_source_table(replica1, "source", False) test_attach_partition_using_copy/test.py:161: _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ test_attach_partition_using_copy/test.py:40: in create_source_table node.query_with_retry( helpers/cluster.py:3713: in query_with_retry result = self.query( helpers/cluster.py:3678: in query return self.client.query( helpers/client.py:39: in wrap return func(self, *args, **kwargs) helpers/client.py:79: in query ).get_answer() helpers/client.py:230: in get_answer self.process.wait(timeout=DEFAULT_QUERY_TIMEOUT) /usr/lib/python3.10/subprocess.py:1209: in wait return self._wait(timeout=timeout) _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ _ self = timeout = 600 def _wait(self, timeout): """Internal implementation of wait() on POSIX.""" if self.returncode is not None: return self.returncode if timeout is not None: endtime = _time() + timeout # Enter a busy loop if we have a timeout. This busy loop was # cribbed from Lib/threading.py in Thread.wait() at r71065. delay = 0.0005 # 500 us -> initial delay of 1 ms while True: if self._waitpid_lock.acquire(False): try: if self.returncode is not None: break # Another thread waited. (pid, sts) = self._try_wait(os.WNOHANG) assert pid == self.pid or pid == 0 if pid == self.pid: self._handle_exitstatus(sts) break finally: self._waitpid_lock.release() remaining = self._remaining_time(endtime) if remaining <= 0: raise TimeoutExpired(self.args, timeout) delay = min(delay * 2, remaining, .05) > time.sleep(delay) E Failed: Timeout >900.0s /usr/lib/python3.10/subprocess.py:1953: Failed ----------------------------- Captured stderr call ----------------------------- Executing query DROP TABLE IF EXISTS source SYNC on replica1 Executing query DROP TABLE IF EXISTS destination SYNC on replica1 Executing query DROP TABLE IF EXISTS source SYNC on replica2 Executing query DROP TABLE IF EXISTS destination SYNC on replica2 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 ~~~~~~~~~~~~~~~~~~~~~ Stack of (139747784914496) ~~~~~~~~~~~~~~~~~~~~~ File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 411, in _perform_spawn reply.run() File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 341, in run self._result = func(*args, **kwargs) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 1160, in _thread_receiver msg = Message.from_io(io) File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 567, in from_io header = io.read(9) # type 1, channel 4, payload 4 File "/usr/local/lib/python3.10/dist-packages/execnet/gateway_base.py", line 534, in read data = self._read(numbytes - len(buf)) ------------------------------ Captured log call ------------------------------- 2025-04-01 22:44:49 [ 694 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica1 (cluster.py:3677, query) 2025-04-01 22:45:04 [ 694 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica1 (cluster.py:3677, query) 2025-04-01 22:45:04 [ 694 ] DEBUG : Executing query DROP TABLE IF EXISTS source SYNC on replica2 (cluster.py:3677, query) 2025-04-01 22:45:04 [ 694 ] DEBUG : Executing query DROP TABLE IF EXISTS destination SYNC on replica2 (cluster.py:3677, query) 2025-04-01 22:45:04 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:46:02 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:46:57 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:47:52 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:48:49 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:49:47 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:50:41 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:51:37 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:52:34 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:53:31 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:54:26 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:55:21 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:56:18 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:57:16 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:58:11 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) 2025-04-01 22:59:06 [ 694 ] DEBUG : Executing query ATTACH TABLE source UUID 'cf712b4f-2ca8-435c-ac23-c4393efe52f7' ( price UInt32, date Date, postcode1 LowCardinality(String), postcode2 LowCardinality(String), type Enum8('other' = 0, 'terraced' = 1, 'semi-detached' = 2, 'detached' = 3, 'flat' = 4), is_new UInt8, duration Enum8('unknown' = 0, 'freehold' = 1, 'leasehold' = 2), addr1 String, addr2 String, street LowCardinality(String), locality LowCardinality(String), town LowCardinality(String), district LowCardinality(String), county LowCardinality(String) ) ENGINE = MergeTree() ORDER BY (postcode1, postcode2, addr1, addr2) SETTINGS disk = disk(type = web, endpoint = 'https://raw.githubusercontent.com/ClickHouse/web-tables-demo/main/web/') on replica1 (cluster.py:3677, query) --------------------------- Captured stderr teardown --------------------------- Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml stop --timeout 20] Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopped Command:[bash -c [ -f /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[bash -c [ -f /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml down --volumes] Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Removing Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Removing Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Removed Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Removed Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopping Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Removing Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Removing Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopped Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Removing Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Removed Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Removed Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Removed Stderr: Network roottestattachpartitionusingcopy-gw6_default Removing Stderr: Network roottestattachpartitionusingcopy-gw6_default Removed Cleanup called Docker networks for project roottestattachpartitionusingcopy-gw6 are NETWORK ID NAME DRIVER SCOPE Docker containers for project roottestattachpartitionusingcopy-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES Docker volumes for project roottestattachpartitionusingcopy-gw6 are DRIVER VOLUME NAME Command:[docker container list --all --filter name='^/roottestattachpartitionusingcopy-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] Unstopped containers: {} No running containers for project: roottestattachpartitionusingcopy-gw6 Trying to prune unused networks... Trying to prune unused images... Command:[docker image prune -f] Stdout:Total reclaimed space: 0B Images pruned Trying to prune unused volumes... Command:[docker volume ls | wc -l] Stdout:1 Volumes pruned: 1 ---------------------------- Captured log teardown ----------------------------- 2025-04-01 22:59:50 [ 694 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml stop --timeout 20] (cluster.py:122, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopping (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopping (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopped (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopped (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopping (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopping (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopping (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopped (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopped (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopped (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:122, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Command:[bash -c [ -f /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/logs/stderr.log ] && zgrep -aH "==================" /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/logs/stderr.log* | ( [ -z "" ] && cat || grep -v "$" ) || true] (cluster.py:122, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Command:[docker compose --env-file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/.env --project-name roottestattachpartitionusingcopy-gw6 --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica1/docker-compose.yml --file /ClickHouse/tests/integration/helpers/../../../tests/integration/compose/docker_compose_keeper.yml --file /ClickHouse/tests/integration/test_attach_partition_using_copy/_instances-0-gw6/replica2/docker-compose.yml down --volumes] (cluster.py:122, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopping (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopping (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Stopped (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Removing (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Stopped (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Removing (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica1-1 Removed (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-replica2-1 Removed (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopping (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopping (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopping (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Stopped (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Removing (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Stopped (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Removing (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Stopped (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Removing (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo3-1 Removed (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo2-1 Removed (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Container roottestattachpartitionusingcopy-gw6-zoo1-1 Removed (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy-gw6_default Removing (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stderr: Network roottestattachpartitionusingcopy-gw6_default Removed (cluster.py:148, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Cleanup called (cluster.py:894, cleanup) 2025-04-01 22:59:59 [ 694 ] DEBUG : Docker networks for project roottestattachpartitionusingcopy-gw6 are NETWORK ID NAME DRIVER SCOPE (cluster.py:873, print_all_docker_pieces) 2025-04-01 22:59:59 [ 694 ] DEBUG : Docker containers for project roottestattachpartitionusingcopy-gw6 are CONTAINER ID IMAGE COMMAND CREATED STATUS PORTS NAMES (cluster.py:881, print_all_docker_pieces) 2025-04-01 22:59:59 [ 694 ] DEBUG : Docker volumes for project roottestattachpartitionusingcopy-gw6 are DRIVER VOLUME NAME (cluster.py:889, print_all_docker_pieces) 2025-04-01 22:59:59 [ 694 ] DEBUG : Command:[docker container list --all --filter name='^/roottestattachpartitionusingcopy-gw6-.*-1$' --format '{{.ID}}:{{.Names}}'] (cluster.py:122, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Unstopped containers: {} (cluster.py:908, cleanup) 2025-04-01 22:59:59 [ 694 ] DEBUG : No running containers for project: roottestattachpartitionusingcopy-gw6 (cluster.py:922, cleanup) 2025-04-01 22:59:59 [ 694 ] DEBUG : Trying to prune unused networks... (cluster.py:928, cleanup) 2025-04-01 22:59:59 [ 694 ] DEBUG : Trying to prune unused images... (cluster.py:944, cleanup) 2025-04-01 22:59:59 [ 694 ] DEBUG : Command:[docker image prune -f] (cluster.py:122, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stdout:Total reclaimed space: 0B (cluster.py:146, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Images pruned (cluster.py:947, cleanup) 2025-04-01 22:59:59 [ 694 ] DEBUG : Trying to prune unused volumes... (cluster.py:953, cleanup) 2025-04-01 22:59:59 [ 694 ] DEBUG : Command:[docker volume ls | wc -l] (cluster.py:122, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Stdout:1 (cluster.py:146, run_and_check) 2025-04-01 22:59:59 [ 694 ] DEBUG : Volumes pruned: 1 (cluster.py:958, cleanup) ============================== slowest durations =============================== 900.00s call test_attach_partition_using_copy/test.py::test_not_work_on_different_disk 900.00s call test_attach_partition_using_copy/test.py::test_only_destination_replicated 900.00s call test_attach_partition_using_copy/test.py::test_both_mergetree 878.37s call test_attach_partition_using_copy/test.py::test_all_replicated 214.98s call test_async_insert_memory/test.py::test_memory_usage 177.74s setup test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication 86.80s call test_ddl_worker_replicas/test.py::test_ddl_worker_replicas 53.14s call test_dictionaries_dependency/test.py::test_no_lazy_load 48.13s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct] 47.75s call test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache] 46.54s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct] 45.32s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache] 44.89s call test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache] 43.10s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed] 42.97s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat] 42.68s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache] 41.88s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed] 24.24s setup test_cluster_all_replicas/test.py::test_cluster 22.74s setup test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] 22.59s call test_concurrent_queries_restriction_by_query_kind/test.py::test_select 22.04s call test_disk_access_storage/test.py::test_alter 22.01s teardown test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation 21.91s call test_config_xml_main/test.py::test_xml_main_conf 21.63s setup test_attach_partition_using_copy/test.py::test_all_replicated 21.57s setup test_always_fetch_merged/test.py::test_replica_always_download 21.02s call test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1] 20.80s setup test_compression_nested_columns/test.py::test_nested_compression_codec 19.84s call test_config_yaml_main/test.py::test_yaml_main_conf 19.75s call test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup 19.56s setup test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0] 19.29s teardown test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication 19.26s call test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0] 18.92s call test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed] 18.79s call test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0] 18.75s setup test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field 18.37s setup test_ddl_worker_replicas/test.py::test_ddl_worker_replicas 18.30s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache] 18.06s call test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1] 17.19s setup test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation 16.92s setup test_backup_restore/test.py::test_attach_partition 16.64s setup test_default_role/test.py::test_alter_user 16.33s setup test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed] 15.16s call test_config_decryption/test_wrong_settings.py::test_invalid_chars 14.54s setup test_accept_invalid_certificate/test.py::test_accept 14.52s call test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0] 14.51s setup test_disk_access_storage/test.py::test_alter 13.96s call test_config_decryption/test_wrong_settings.py::test_no_encryption_key 13.84s setup test_create_query_constraints/test.py::test_create_query_const_constraints 13.74s call test_config_decryption/test_wrong_settings.py::test_wrong_method 12.75s call test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1] 12.73s teardown test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica 12.61s call test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1] 12.57s setup test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica 12.38s call test_always_fetch_merged/test.py::test_replica_always_download 12.25s call test_config_decryption/test_wrong_settings.py::test_subnodes 11.95s teardown test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] 11.55s teardown test_backup_restore_on_cluster/test.py::test_tables_dependency 10.99s call test_concurrent_queries_restriction_by_query_kind/test.py::test_insert 10.52s call test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0] 10.39s setup test_custom_settings/test.py::test_custom_settings 9.78s setup test_concurrent_queries_restriction_by_query_kind/test.py::test_insert 9.60s teardown test_attach_partition_using_copy/test.py::test_only_destination_replicated 9.41s call test_backup_restore_on_cluster/test.py::test_replicated_database_async 8.30s call test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] 8.22s call test_backup_restore_on_cluster/test.py::test_tables_dependency 8.12s teardown test_dictionaries_dependency/test.py::test_no_lazy_load 8.01s call test_backup_restore/test.py::test_replace_partition 7.74s call test_backup_restore/test.py::test_attach_partition 7.30s setup test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache] 7.22s call test_backup_restore_on_cluster/test.py::test_required_privileges 6.84s call test_backup_restore_on_cluster/test.py::test_keeper_value_max_size 6.84s teardown test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1] 6.79s call test_backup_restore_on_cluster/test.py::test_replicated_database 6.56s teardown test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1] 6.51s teardown test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1] 6.51s call test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 6.39s call test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica 6.34s call test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] 6.33s call test_backup_restore/test.py::test_restore 6.33s teardown test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1] 6.33s call test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] 6.31s teardown test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0] 6.28s call test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] 6.15s call test_backup_restore_on_cluster/test.py::test_mutation 6.12s teardown test_always_fetch_merged/test.py::test_replica_always_download 5.92s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def 5.91s teardown test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0] 5.87s teardown test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0] 5.85s teardown test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0] 5.80s teardown test_async_insert_memory/test.py::test_memory_usage 5.76s teardown test_concurrent_queries_restriction_by_query_kind/test.py::test_select 5.55s call test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def 5.48s teardown test_ddl_worker_replicas/test.py::test_ddl_worker_replicas 5.41s call test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] 5.28s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed] 5.21s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert 5.21s call test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field 5.16s call test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] 5.14s call test_backup_restore_on_cluster/test.py::test_file_deduplication 5.06s call test_backup_restore_on_cluster/test.py::test_replicated_table 4.92s call test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation 4.89s teardown test_compression_nested_columns/test.py::test_nested_compression_codec 4.86s teardown test_custom_settings/test.py::test_illformed_setting 4.85s teardown test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field 4.73s teardown test_accept_invalid_certificate/test.py::test_strict_reject_with_config 4.72s teardown test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache] 4.68s call test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica 4.66s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge 4.59s call test_backup_restore_on_cluster/test.py::test_system_users 4.59s call test_backup_restore_on_cluster/test.py::test_system_functions 4.52s setup test_async_insert_memory/test.py::test_memory_usage 4.48s call test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] 4.45s call test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath 4.44s call test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] 4.33s call test_backup_restore_on_cluster/test.py::test_projection 4.32s call test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster 4.10s call test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty 4.08s teardown test_backup_restore/test.py::test_restore 3.96s call test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] 3.71s call test_default_role/test.py::test_set_default_roles 3.68s teardown test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache] 3.57s call test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache] 3.56s call test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication 3.49s call test_compression_nested_columns/test.py::test_nested_compression_codec 3.45s call test_create_query_constraints/test.py::test_create_query_minmax_constraints 3.25s teardown test_create_query_constraints/test.py::test_create_query_minmax_constraints 3.24s call test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache] 3.24s call test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache] 3.16s call test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes 3.07s call test_backup_restore_on_cluster/test.py::test_empty_replicated_table 3.01s call test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster 2.92s call test_default_role/test.py::test_alter_user 2.79s teardown test_default_role/test.py::test_wrong_set_default_role 2.61s call test_create_query_constraints/test.py::test_create_query_const_constraints 2.37s teardown test_backup_restore_on_cluster/test.py::test_file_deduplication 2.33s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_async 2.25s teardown test_backup_restore_on_cluster/test.py::test_replicated_database 2.22s call test_custom_settings/test.py::test_custom_settings 2.14s teardown test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty 2.13s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] 2.07s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def 2.04s teardown test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] 2.03s teardown test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] 2.03s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] 1.99s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath 1.98s teardown test_backup_restore_on_cluster/test.py::test_get_error_from_other_host 1.98s teardown test_backup_restore_on_cluster/test.py::test_replicated_table 1.98s teardown test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup 1.98s teardown test_backup_restore_on_cluster/test.py::test_mutation 1.93s teardown test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def 1.93s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster 1.93s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster 1.89s teardown test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica 1.88s teardown test_backup_restore_on_cluster/test.py::test_projection 1.88s teardown test_backup_restore_on_cluster/test.py::test_system_users 1.88s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert 1.85s teardown test_backup_restore_on_cluster/test.py::test_required_privileges 1.83s teardown test_backup_restore_on_cluster/test.py::test_system_functions 1.83s teardown test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 1.83s teardown test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes 1.81s teardown test_backup_restore_on_cluster/test.py::test_keeper_value_max_size 1.78s teardown test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] 1.78s teardown test_backup_restore_on_cluster/test.py::test_empty_replicated_table 1.78s teardown test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge 1.43s call test_default_role/test.py::test_wrong_set_default_role 1.36s call test_cluster_all_replicas/test.py::test_cluster 1.32s call test_backup_restore_on_cluster/test.py::test_get_error_from_other_host 1.10s call test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed] 0.95s teardown test_disk_access_storage/test.py::test_alter 0.90s call test_cluster_all_replicas/test.py::test_global_in 0.85s call test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed] 0.80s call test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed] 0.75s call test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat] 0.70s call test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed] 0.61s call test_custom_settings/test.py::test_illformed_setting 0.42s call test_accept_invalid_certificate/test.py::test_accept 0.37s call test_accept_invalid_certificate/test.py::test_connection_accept 0.27s setup test_default_role/test.py::test_set_default_roles 0.27s call test_accept_invalid_certificate/test.py::test_strict_reject_with_config 0.22s call test_accept_invalid_certificate/test.py::test_strict_connection_reject 0.22s call test_accept_invalid_certificate/test.py::test_default 0.22s setup test_default_role/test.py::test_wrong_set_default_role 0.22s call test_accept_invalid_certificate/test.py::test_strict_reject 0.04s setup test_config_decryption/test_wrong_settings.py::test_invalid_chars 0.00s setup test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica 0.00s teardown test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache] 0.00s setup test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] 0.00s teardown test_config_decryption/test_wrong_settings.py::test_subnodes 0.00s setup test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_async 0.00s teardown test_config_yaml_main/test.py::test_yaml_main_conf 0.00s setup test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct] 0.00s setup test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0] 0.00s setup test_backup_restore_on_cluster/test.py::test_mutation 0.00s setup test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1] 0.00s setup test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1] 0.00s setup test_config_xml_main/test.py::test_xml_main_conf 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath 0.00s setup test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] 0.00s teardown test_default_role/test.py::test_alter_user 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge 0.00s setup test_config_yaml_main/test.py::test_yaml_main_conf 0.00s setup test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def 0.00s setup test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1] 0.00s setup test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 0.00s setup test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat] 0.00s setup test_backup_restore_on_cluster/test.py::test_file_deduplication 0.00s setup test_create_query_constraints/test.py::test_create_query_minmax_constraints 0.00s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat] 0.00s setup test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_database 0.00s setup test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster 0.00s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed] 0.00s setup test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache] 0.00s setup test_backup_restore_on_cluster/test.py::test_projection 0.00s teardown test_cluster_all_replicas/test.py::test_cluster 0.00s setup test_backup_restore_on_cluster/test.py::test_system_users 0.00s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct] 0.00s setup test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed] 0.00s setup test_backup_restore_on_cluster/test.py::test_empty_replicated_table 0.00s setup test_backup_restore/test.py::test_restore 0.00s setup test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed] 0.00s setup test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed] 0.00s setup test_backup_restore_on_cluster/test.py::test_required_privileges 0.00s setup test_dictionaries_dependency/test.py::test_no_lazy_load 0.00s setup test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed] 0.00s teardown test_attach_partition_using_copy/test.py::test_all_replicated 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster 0.00s teardown test_concurrent_queries_restriction_by_query_kind/test.py::test_insert 0.00s setup test_backup_restore_on_cluster/test.py::test_system_functions 0.00s teardown test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed] 0.00s setup test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] 0.00s setup test_backup_restore_on_cluster/test.py::test_tables_dependency 0.00s setup test_accept_invalid_certificate/test.py::test_strict_reject 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed] 0.00s setup test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct] 0.00s setup test_backup_restore/test.py::test_replace_partition 0.00s setup test_backup_restore_on_cluster/test.py::test_keeper_value_max_size 0.00s teardown test_config_xml_main/test.py::test_xml_main_conf 0.00s setup test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat] 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache] 0.00s setup test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert 0.00s teardown test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed] 0.00s setup test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache] 0.00s setup test_custom_settings/test.py::test_illformed_setting 0.00s setup test_accept_invalid_certificate/test.py::test_connection_accept 0.00s setup test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes 0.00s teardown test_custom_settings/test.py::test_custom_settings 0.00s setup test_concurrent_queries_restriction_by_query_kind/test.py::test_select 0.00s setup test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache] 0.00s teardown test_create_query_constraints/test.py::test_create_query_const_constraints 0.00s setup test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] 0.00s teardown test_backup_restore/test.py::test_attach_partition 0.00s setup test_backup_restore_on_cluster/test.py::test_get_error_from_other_host 0.00s teardown test_default_role/test.py::test_set_default_roles 0.00s teardown test_cluster_all_replicas/test.py::test_global_in 0.00s teardown test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed] 0.00s teardown test_attach_partition_using_copy/test.py::test_both_mergetree 0.00s setup test_config_decryption/test_wrong_settings.py::test_no_encryption_key 0.00s setup test_cluster_all_replicas/test.py::test_global_in 0.00s teardown test_config_decryption/test_wrong_settings.py::test_invalid_chars 0.00s setup test_accept_invalid_certificate/test.py::test_strict_reject_with_config 0.00s teardown test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed] 0.00s teardown test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed] 0.00s teardown test_accept_invalid_certificate/test.py::test_accept 0.00s setup test_config_decryption/test_wrong_settings.py::test_wrong_method 0.00s teardown test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat] 0.00s teardown test_attach_partition_using_copy/test.py::test_not_work_on_different_disk 0.00s teardown test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] 0.00s teardown test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] 0.00s teardown test_config_decryption/test_wrong_settings.py::test_no_encryption_key 0.00s teardown test_config_decryption/test_wrong_settings.py::test_wrong_method 0.00s setup test_attach_partition_using_copy/test.py::test_not_work_on_different_disk 0.00s setup test_accept_invalid_certificate/test.py::test_default 0.00s teardown test_backup_restore/test.py::test_replace_partition 0.00s setup test_accept_invalid_certificate/test.py::test_strict_connection_reject 0.00s teardown test_accept_invalid_certificate/test.py::test_strict_connection_reject 0.00s setup test_config_decryption/test_wrong_settings.py::test_subnodes 0.00s setup test_attach_partition_using_copy/test.py::test_both_mergetree 0.00s teardown test_accept_invalid_certificate/test.py::test_strict_reject 0.00s teardown test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache] 0.00s setup test_attach_partition_using_copy/test.py::test_only_destination_replicated 0.00s teardown test_accept_invalid_certificate/test.py::test_connection_accept 0.00s teardown test_accept_invalid_certificate/test.py::test_default 0.00s teardown test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache] =========================== short test summary info ============================ FAILED test_attach_partition_using_copy/test.py::test_all_replicated - Failed... FAILED test_attach_partition_using_copy/test.py::test_both_mergetree - Failed... FAILED test_attach_partition_using_copy/test.py::test_not_work_on_different_disk FAILED test_attach_partition_using_copy/test.py::test_only_destination_replicated PASSED test_accept_invalid_certificate/test.py::test_accept PASSED test_config_decryption/test_wrong_settings.py::test_invalid_chars PASSED test_accept_invalid_certificate/test.py::test_connection_accept PASSED test_accept_invalid_certificate/test.py::test_default PASSED test_accept_invalid_certificate/test.py::test_strict_connection_reject PASSED test_accept_invalid_certificate/test.py::test_strict_reject PASSED test_accept_invalid_certificate/test.py::test_strict_reject_with_config PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed] PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed] PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat] PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed] PASSED test_default_role/test.py::test_alter_user PASSED test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed] PASSED test_default_role/test.py::test_set_default_roles PASSED test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache] PASSED test_backup_restore/test.py::test_attach_partition PASSED test_default_role/test.py::test_wrong_set_default_role PASSED test_cluster_all_replicas/test.py::test_cluster PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False] PASSED test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache] PASSED test_config_decryption/test_wrong_settings.py::test_no_encryption_key PASSED test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache] PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes] PASSED test_backup_restore/test.py::test_replace_partition PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True] PASSED test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes] PASSED test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0] PASSED test_backup_restore/test.py::test_restore PASSED test_cluster_all_replicas/test.py::test_global_in PASSED test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True] PASSED test_config_decryption/test_wrong_settings.py::test_subnodes PASSED test_concurrent_queries_restriction_by_query_kind/test.py::test_insert PASSED test_create_query_constraints/test.py::test_create_query_const_constraints PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes] PASSED test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica PASSED test_create_query_constraints/test.py::test_create_query_minmax_constraints PASSED test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes PASSED test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes] PASSED test_config_decryption/test_wrong_settings.py::test_wrong_method PASSED test_backup_restore_on_cluster/test.py::test_empty_replicated_table PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache] PASSED test_backup_restore_on_cluster/test.py::test_file_deduplication PASSED test_concurrent_queries_restriction_by_query_kind/test.py::test_select PASSED test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1] PASSED test_backup_restore_on_cluster/test.py::test_get_error_from_other_host PASSED test_backup_restore_on_cluster/test.py::test_keeper_value_max_size PASSED test_always_fetch_merged/test.py::test_replica_always_download PASSED test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field PASSED test_custom_settings/test.py::test_custom_settings PASSED test_custom_settings/test.py::test_illformed_setting PASSED test_backup_restore_on_cluster/test.py::test_mutation PASSED test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0] PASSED test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache] PASSED test_config_xml_main/test.py::test_xml_main_conf PASSED test_backup_restore_on_cluster/test.py::test_projection PASSED test_backup_restore_on_cluster/test.py::test_replicated_database PASSED test_config_yaml_main/test.py::test_yaml_main_conf PASSED test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct] PASSED test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_async PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts PASSED test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database] PASSED test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid] PASSED test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table PASSED test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def PASSED test_compression_nested_columns/test.py::test_nested_compression_codec PASSED test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed] PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge PASSED test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath PASSED test_backup_restore_on_cluster/test.py::test_required_privileges PASSED test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1] PASSED test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache] PASSED test_backup_restore_on_cluster/test.py::test_system_functions PASSED test_backup_restore_on_cluster/test.py::test_system_users PASSED test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def PASSED test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty PASSED test_backup_restore_on_cluster/test.py::test_tables_dependency PASSED test_dictionaries_dependency/test.py::test_no_lazy_load PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct] PASSED test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication PASSED test_async_insert_memory/test.py::test_memory_usage PASSED test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat] PASSED test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed] PASSED test_ddl_worker_replicas/test.py::test_ddl_worker_replicas PASSED test_disk_access_storage/test.py::test_alter ================== 4 failed, 96 passed in 3612.60s (1:00:12) =================== Traceback (most recent call last): File "/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration/./runner", line 528, in subprocess.check_call(cmd, shell=True) File "/usr/lib/python3.10/subprocess.py", line 369, in check_call raise CalledProcessError(retcode, cmd) subprocess.CalledProcessError: Command 'docker run --rm --name clickhouse_integration_tests_bvh7ap --privileged --dns-search='.' --memory=30709035008 --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-odbc-bridge:/clickhouse-odbc-bridge --volume=/home/ubuntu/_work/_temp/test/build/clickhouse:/clickhouse --volume=/home/ubuntu/_work/_temp/test/build/clickhouse-library-bridge:/clickhouse-library-bridge --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/programs/server:/clickhouse-config --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/tests/integration:/ClickHouse/tests/integration --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/backupview:/ClickHouse/utils/backupview --volume=/home/ubuntu/_work/ClickHouse/ClickHouse/utils/grpc-client/pb2:/ClickHouse/utils/grpc-client/pb2 --volume=/run:/run/host:ro --volume=clickhouse_integration_tests_volume:/var/lib/docker -e DOCKER_DOTNET_CLIENT_TAG=11de0b29a15d -e DOCKER_HELPER_TAG=5dc43a6382f0 -e DOCKER_BASE_TAG=6712d5cc610d -e DOCKER_KERBEROS_KDC_TAG=9391ecdee8d7 -e DOCKER_MYSQL_GOLANG_CLIENT_TAG=9bec2a638e6e -e DOCKER_MYSQL_JAVA_CLIENT_TAG=766bff31cfe4 -e DOCKER_MYSQL_JS_CLIENT_TAG=41ba7c2ec2a1 -e DOCKER_MYSQL_PHP_CLIENT_TAG=88be89c1e3b6 -e DOCKER_NGINX_DAV_TAG=b55ac9cd7519 -e DOCKER_POSTGRESQL_JAVA_CLIENT_TAG=a4eff5c7f4d6 -e DOCKER_PYTHON_BOTTLE_TAG=caad4729259e -e DOCKER_CLIENT_TIMEOUT=300 -e COMPOSE_HTTP_TIMEOUT=600 -e PYTHONUNBUFFERED=1 -e PYTEST_ADDOPTS="--dist=loadfile -n 10 -rfEps --run-id=0 --color=no --durations=0 test_accept_invalid_certificate/test.py::test_accept test_accept_invalid_certificate/test.py::test_connection_accept test_accept_invalid_certificate/test.py::test_default test_accept_invalid_certificate/test.py::test_strict_connection_reject test_accept_invalid_certificate/test.py::test_strict_reject test_accept_invalid_certificate/test.py::test_strict_reject_with_config test_always_fetch_merged/test.py::test_replica_always_download test_async_insert_memory/test.py::test_memory_usage test_asynchronous_metric_log_table/test.py::test_event_time_microseconds_field test_attach_partition_using_copy/test.py::test_all_replicated test_attach_partition_using_copy/test.py::test_both_mergetree test_attach_partition_using_copy/test.py::test_not_work_on_different_disk test_attach_partition_using_copy/test.py::test_only_destination_replicated test_azure_blob_storage_zero_copy_replication/test.py::test_zero_copy_replication test_backup_restore/test.py::test_attach_partition test_backup_restore/test.py::test_replace_partition test_backup_restore/test.py::test_restore 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-False]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[http-True]' 'test_backup_restore_on_cluster/test.py::test_async_backups_to_same_destination[native-True]' test_backup_restore_on_cluster/test.py::test_backup_restore_on_single_replica test_backup_restore_on_cluster/test.py::test_different_tables_on_nodes test_backup_restore_on_cluster/test.py::test_empty_replicated_table test_backup_restore_on_cluster/test.py::test_file_deduplication test_backup_restore_on_cluster/test.py::test_get_error_from_other_host test_backup_restore_on_cluster/test.py::test_keeper_value_max_size test_backup_restore_on_cluster/test.py::test_mutation test_backup_restore_on_cluster/test.py::test_projection test_backup_restore_on_cluster/test.py::test_replicated_database test_backup_restore_on_cluster/test.py::test_replicated_database_async test_backup_restore_on_cluster/test.py::test_replicated_database_compare_parts 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[database]' 'test_backup_restore_on_cluster/test.py::test_replicated_database_with_special_macro_in_zk_path[uuid]' test_backup_restore_on_cluster/test.py::test_replicated_table test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_bigger_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_restored_into_smaller_cluster test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_def test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_insert test_backup_restore_on_cluster/test.py::test_replicated_table_with_not_synced_merge test_backup_restore_on_cluster/test.py::test_replicated_table_with_uuid_in_zkpath test_backup_restore_on_cluster/test.py::test_required_privileges test_backup_restore_on_cluster/test.py::test_shutdown_waits_for_backup test_backup_restore_on_cluster/test.py::test_system_functions test_backup_restore_on_cluster/test.py::test_system_users test_backup_restore_on_cluster/test.py::test_table_in_replicated_database_with_not_synced_def test_backup_restore_on_cluster/test.py::test_table_with_parts_in_queue_considered_non_empty test_backup_restore_on_cluster/test.py::test_tables_dependency test_cluster_all_replicas/test.py::test_cluster 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_error_on_unavailable_replica[two_shards_three_nodes]' test_cluster_all_replicas/test.py::test_global_in 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[one_shard_three_nodes]' 'test_cluster_all_replicas/test.py::test_skip_unavailable_replica[two_shards_three_nodes]' test_compression_nested_columns/test.py::test_nested_compression_codec test_concurrent_queries_restriction_by_query_kind/test.py::test_insert test_concurrent_queries_restriction_by_query_kind/test.py::test_select test_config_decryption/test_wrong_settings.py::test_invalid_chars test_config_decryption/test_wrong_settings.py::test_no_encryption_key test_config_decryption/test_wrong_settings.py::test_subnodes test_config_decryption/test_wrong_settings.py::test_wrong_method test_config_xml_main/test.py::test_xml_main_conf test_config_yaml_main/test.py::test_yaml_main_conf test_create_query_constraints/test.py::test_create_query_const_constraints test_create_query_constraints/test.py::test_create_query_minmax_constraints test_custom_settings/test.py::test_custom_settings test_custom_settings/test.py::test_illformed_setting test_ddl_config_hostname/test.py::test_ddl_queue_delete_add_replica test_ddl_worker_replicas/test.py::test_ddl_worker_replicas test_default_role/test.py::test_alter_user test_default_role/test.py::test_set_default_roles test_default_role/test.py::test_wrong_set_default_role 'test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_executable_cache.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_cache]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_direct]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_complex[complex_key_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_ranged[range_hashed]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[cache]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[direct]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[flat]' 'test_dictionaries_all_layouts_separate_sources/test_http.py::test_simple[hashed]' 'test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node0]' 'test_dictionaries_dependency/test.py::test_dependency_via_dictionary_database[node1]' 'test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node0]' 'test_dictionaries_dependency/test.py::test_dependency_via_explicit_table[node1]' 'test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node0]' 'test_dictionaries_dependency/test.py::test_dependency_via_implicit_table[node1]' 'test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node0]' 'test_dictionaries_dependency/test.py::test_dependent_dict_table_distr[node1]' test_dictionaries_dependency/test.py::test_no_lazy_load 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_integers_key_hashed]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_complex_mixed_key_hashed]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_flat]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_hashed]' 'test_dictionaries_select_all/test.py::test_select_all[clickhouse_range_hashed]' 'test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_cache]' 'test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_integers_key_cache]' 'test_dictionaries_select_all/test.py::test_select_all_from_cached[clickhouse_complex_mixed_key_cache]' test_disable_insertion_and_mutation/test.py::test_disable_insertion_and_mutation test_disk_access_storage/test.py::test_alter -vvv" altinityinfra/integration-tests-runner:cd6390247eca ' returned non-zero exit status 1.